Journal article

Evaluating large language models for multilingual vulnerability detection at dual granularities

H Shu, M Fu, J Yu, D Wang, C Tantithamthavorn, J Chen, Y Kamei

Empirical Software Engineering | Published : 2026

Abstract

Various deep learning-based approaches utilizing pre-trained language models (PLMs) have been proposed for automated vulnerability detection. With recent advancements in large language models (LLMs), several studies have begun exploring their application to vulnerability detection tasks. However, existing studies primarily focus on specific programming languages (e.g., C/C++) and function-level detection, leaving the strengths and weaknesses of PLMs and LLMs in multilingual and multi-granularity scenarios largely unexplored. To bridge this gap, we conduct a comprehensive fine-grained empirical study evaluating the effectiveness of state-of-the-art PLMs and LLMs for multilingual vulnerability..

View full abstract

University of Melbourne Researchers

Grants

Awarded by Key Technologies Research and Development Program


Awarded by National Natural Science Foundation of China


Awarded by Japan Science and Technology Corporation


Awarded by Japan Society for the Promotion of Science London